16 research outputs found

    Credit Card Fraud Detection Using Machine Learning As Data Mining Technique

    Get PDF
    The rapid participation in online based transactional activities raises the fraudulent cases all over the world and causes tremendous losses to the individuals and financial industry. Although there are many criminal activities occurring in financial industry, credit card fraudulent activities are among the most prevalent and worried about by online customers. Thus, countering the fraud activities through data mining and machine learning is one of the prominent approaches introduced by scholars intending to prevent the losses caused by these illegal acts. Primarily, data mining techniques were employed to study the patterns and characteristics of suspicious and non-suspicious transactions based on normalized and anomalies data. On the other hand, machine learning (ML) techniques were employed to predict the suspicious and non-suspicious transactions automatically by using classifiers. Therefore, the combination of machine learning and data mining techniques were able to identify the genuine and non-genuine transactions by learning the patterns of the data. This paper discusses the supervised based classification using Bayesian network classifiers namely K2, Tree Augmented Naïve Bayes (TAN), and Naïve Bayes, logistics and J48 classifiers. After preprocessing the dataset using normalization and Principal Component Analysis, all the classifiers achieved more than 95.0% accuracy compared to results attained before preprocessing the dataset

    Math E-Tutor: Towards A Better Self-Learning Environment.

    Get PDF
    Math E-Tutor is a web-based tutoring system which aims to create a virtual teaching and learning environment that suits today's trend. This system focuses on the mathematical syllabus of lower secondary school which uses multiple choice questions (MCQ) as the mode of evaluation, known as Penilaian Menengah Rendah (PMR) in Malaysia

    Application Of Exact String Matching Algorithms Towards SMILES Representation Of Chemical Structure.

    Get PDF
    Bioinformatics and Cheminformatics use computer as disciplines providing tools for acquisition, storage, processing, analysis, integrate data and for the development of potential applications of biological and chemical data. A chemical database is one of the databases that exclusively designed to store chemical information

    DeepNC: a framework for drug-target interaction prediction with graph neural networks

    Get PDF
    The exploration of drug-target interactions (DTI) is an essential stage in the drug development pipeline. Thanks to the assistance of computational models, notably in the deep learning approach, scientists have been able to shorten the time spent on this stage. Widely practiced deep learning algorithms such as convolutional neural networks and recurrent neural networks are commonly employed in DTI prediction projects. However, they can hardly utilize the natural graph structure of molecular inputs. For that reason, a graph neural network (GNN) is an applicable choice for learning the chemical and structural characteristics of molecules when it represents molecular compounds as graphs and learns the compound features from those graphs. In an effort to construct an advanced deep learning-based model for DTI prediction, we propose Deep Neural Computation (DeepNC), which is a framework utilizing three GNN algorithms: Generalized Aggregation Networks (GENConv), Graph Convolutional Networks (GCNConv), and Hypergraph Convolution-Hypergraph Attention (HypergraphConv). In short, our framework learns the features of drugs and targets by the layers of GNN and 1-D convolution network, respectively. Then, representations of the drugs and targets are fed into fully-connected layers to predict the binding affinity values. The models of DeepNC were evaluated on two benchmarked datasets (Davis, Kiba) and one independently proposed dataset (Allergy) to confirm that they are suitable for predicting the binding affinity of drugs and targets. Moreover, compared to the results of baseline methods that worked on the same problem, DeepNC proves to improve the performance in terms of mean square error and concordance index

    Applications of Brain Computer Interface in Present Healthcare Setting

    Get PDF
    Brain-computer interface (BCI) is an innovative method of integrating technology for healthcare. Utilizing BCI technology allows for direct communication and/or control between the brain and an external device, thereby displacing conventional neuromuscular pathways. The primary goal of BCI in healthcare is to repair or reinstate useful function to people who have impairments caused by neuromuscular disorders (e.g., stroke, amyotrophic lateral sclerosis, spinal cord injury, or cerebral palsy). BCI brings with it technical and usability flaws in addition to its benefits. We present an overview of BCI in this chapter, followed by its applications in the medical sector in diagnosis, rehabilitation, and assistive technology. We also discuss BCI’s strengths and limitations, as well as its future direction

    Library advisor - an information management system on standard cell library

    Get PDF
    Integrated Circuit (IC) Design has always been a challenge to designers.There are two approaches that can be used to draft architecture of an integrated circuit which are the old approach of using Custom and a new one using ASIC.ASIC has been preferred by IC designers as it manipulates liberty files to provide suggestion on the architecture design faster than the Custom approach that relies on tailor-made designs. Tailor-made designs are difficult to draft and not flexible nor scalable.In fact, the time required for an IC to be build based on Custom design is longer than the one using ASIC.Furthermore, ASIC works on a higher-level of the IC Design in which liberty libraries can be sought to find suitable cells for the IC.However, these liberty libraries are huge that it is quite impossible to browse through manually for intensive comparisons to be made.This has been the obstacle in the ASIC design.Therefore, we proposed a system that stores these liberty libraries as a database and is able to intelligently suggest possible candidate components for IC design sought by the system user.This is a new and novel system that would benefit all designers in the IC design field

    Security attacks taxonomy on bring your own devices (BYOD) model

    Get PDF
    Mobile devices, specifically smartphones, have become ubiquitous. For this reason, businesses are starting to develop “Bring Your Own Device” policies to allow their employees to use their owned devices in the workplace. BYOD offers many potential advantages: enhanced productivity, increased revenues, reduced mobile costs and IT efficiencies. However, due to emerging attacks and limitations on device resources, it is difficult to trust these devices with access to critical proprietary information. Therefore, in this paper, the potential attacks of BYOD and taxonomy of BYOD attacks are presented. Advanced persistent threat (APT) and malware attack are discussed in depth in this paper. Next, the proposed solution to mitigate the attacks of BYOD is discussed. Lastly, the evaluations of the proposed solutions based on the X. 800 security architecture are presented

    Localizing user experience for mobile application: A case study among USM undergraduates

    Get PDF
    The increasing number of mobile devices consumptions, especially mobile phones and smartphones had caused a growing interest in the user experience research on the mobile platform. However, it is difficult to gain an encompassing understanding of the user experience especially from a localized context.This paper presents our survey finding on the understanding of user experience among USM undergraduates.The research finding identifies that a respondents’ field of study has some influence on their understanding of user experience.Overall, the respondents tend to agree that user experience is subjective and based on the individual’s interaction with an application. The ISO definition of user experience is discussed and compared with the survey results. The comparison highlights that the definition of user experience is similar with the ISO’s definition of user experience

    A Novel Algorithm for Estimating Web Page Ranking in Search Engine Results Pages

    No full text
    Abstract: Search engine optimization (SEO) can make a big improvement in the traffic to a web page. Because search engines keep their main rules of ranking undeclared, it’s important to develop models that can estimate the ranking of a web page in the search engine to be able to optimize web pages to rank higher in the search engine. The available research methodologies used machine learning algorithms to provide solutions for this target with the help of generated datasets by scraping the search engine results pages (SERP) and crawling web pages. Their proposed models suffered from the inability to be updated dynamically if the search engine updated its ranking algorithm, and their input data did not include the diversity of web pages and languages. This research will propose a novel original rank estimation algorithm that’s able to overcome other research challenges, with a set of comparative experiments and complexity analysis. Results will show that the proposed algorithm could achieve higher values of accuracy, precision, and recall. Dataset: For research purpose, the dataset will play two roles, first, it will act the role of search engine result pages (SERP), and second, it will be used to test algorithms and calculate performance measurements. Dataset is consisting of 9930 web pages, aimed to identify search results pages, focusing on the top 3 pages of SERP, with 31 extracted attributes that's related to search engine optimization (SEO). The distribution of examples between class labels was balanced, with changes due to scraping operation issues, but not significantly different, with fractions of 39.9%, 34.6%, and 25.5% for the class labels page1, page2, and page 3. Feature names are: 'Title 1 Length', 'Title 2 Length', 'Meta Description 1 Length', 'Meta Description 2 Length', 'Meta Keywords 1 Length', 'H1-1 Length', 'H1-2 Length', 'H2-1 Length', 'H2-2 Length', 'Size (bytes)', 'Word Count', 'Text Ratio', 'Inlinks', 'Unique Inlinks', 'Unique JS Inlinks', '% of Total', 'Outlinks', 'Unique Outlinks', 'Unique JS Outlinks', 'External Outlinks', 'Unique External Outlinks', 'Unique External JS Outlinks', 'Response Time', 'Status Code', 'Keyword in MetaDescription1', 'Keyword in Title1', 'Keyword in MetaKeywords1', 'Keyword in URL', 'Has LastModified', 'Keyword in Headers', and 'Keyword in Emphasized Text'. The process of dataset generation involved scraping the search engine, extracting URLs for selected keywords, focusing on feature extraction, cleaning and preprocessing, and generating new attributes related to keywords in web pages. It involved also removing missing values, duplicates, and data type conversions to obtain a comprehensive dataset. Keyword selection involves selecting keywords from various categories and considering diversity, including high and low traffic, long-term and short-term keywords, and generic and branded keywords. Apify online tool was used for search engine scraping with default language and US country, resulting in 388 selected keywords with 30 results per keyword. Dataset included extracted SEO features from 9991 web pages using screamingFrog desktop software and Rapidminer desktop software, determining page SEO-friendliness and comparing it to SERP rankings. Dataset cleaning involved removing redundant attributes, removing paid SERP results, replacing missing values, and converting data types. Rapidminer was used for data cleaning and preprocessing, generating new attributes related to keyword usage in web pages.This publication includes datasets, Rapidminer processes, and Python codes

    Comparative Studies on Resampling Techniques in Machine Learning and Deep Learning Models for Drug-Target Interaction Prediction

    No full text
    The prediction of drug-target interactions (DTIs) is a vital step in drug discovery. The success of machine learning and deep learning methods in accurately predicting DTIs plays a huge role in drug discovery. However, when dealing with learning algorithms, the datasets used are usually highly dimensional and extremely imbalanced. To solve this issue, the dataset must be resampled accordingly. In this paper, we have compared several data resampling techniques to overcome class imbalance in machine learning methods as well as to study the effectiveness of deep learning methods in overcoming class imbalance in DTI prediction in terms of binary classification using ten (10) cancer-related activity classes from BindingDB. It is found that the use of Random Undersampling (RUS) in predicting DTIs severely affects the performance of a model, especially when the dataset is highly imbalanced, thus, rendering RUS unreliable. It is also found that SVM-SMOTE can be used as a go-to resampling method when paired with the Random Forest and Gaussian Naïve Bayes classifiers, whereby a high F1 score is recorded for all activity classes that are severely and moderately imbalanced. Additionally, the deep learning method called Multilayer Perceptron recorded high F1 scores for all activity classes even when no resampling method was applied
    corecore